parallel computer - перевод на Английский
Diclib.com
Словарь ChatGPT
Введите слово или словосочетание на любом языке 👆
Язык:

Перевод и анализ слов искусственным интеллектом ChatGPT

На этой странице Вы можете получить подробный анализ слова или словосочетания, произведенный с помощью лучшей на сегодняшний день технологии искусственного интеллекта:

  • как употребляется слово
  • частота употребления
  • используется оно чаще в устной или письменной речи
  • варианты перевода слова
  • примеры употребления (несколько фраз с переводом)
  • этимология

parallel computer - перевод на Английский

PROGRAMMING PARADIGM IN WHICH MANY CALCULATIONS OR THE EXECUTION OF PROCESSES ARE CARRIED OUT SIMULTANEOUSLY
Parallel computer; Parallel processor; Parallel computation; Parallel programming; Parallel Programming; Parallel computers; Concurrent language; Concurrent event; Computer Parallelism; Parallel machine; Concurrent (programming); Parallel architecture; Parallel Computing; Parallelisation; Parallelization; Parallelized; Multicomputer; Parallelism (computing); Parellel computing; Superword Level Parallelism; Parallel programming language; Message-driven parallel programming; Parallel computer hardware; Parallel program; Parallel code; Parallel language; Parallel processing (computing); Multiple processing elements; Parallel execution units; History of parallel computing; Parallel hardware; Parallel processing computer
  • A graphical representation of [[Amdahl's law]]. The speedup of a program from parallelization is limited by how much of the program can be parallelized. For example, if 90% of the program can be parallelized, the theoretical maximum speedup using parallel computing would be 10 times no matter how many processors are used.
  • Beowulf cluster]]
  • Blue Gene/L]] massively parallel [[supercomputer]]
  • The [[Cray-1]] is a vector processor
  • 1=IPC = 1}}).
  • A graphical representation of [[Gustafson's law]]
  • Blue Gene/P]] [[massively parallel]] [[supercomputer]]
  • [[ILLIAC IV]], "the most infamous of supercomputers"<ref name="infamous"/>
  • 1=IPC = 0.2 < 1}}).
  • A logical view of a [[non-uniform memory access]] (NUMA) architecture. Processors in one directory can access that directory's memory with less latency than they can access memory in the other directory's memory.
  • Tesla GPGPU card]]
  • 1=IPC = 2 > 1}}).
  • Taiwania 3 of [[Taiwan]], a parallel supercomputing device that joined [[COVID-19]] research.
Найдено результатов: 2387
parallel computer         

[pærəlelkəm'pju:tə]

общая лексика

параллельный компьютер

вычислительная система с распараллеливанием выполнения операций на множестве арифметических и логических блоков (процессоров). Различают системы с общей и распределённой оперативной памятью, многопроцессорные и многокомпьютерные системы

параллельная ЭВМ

Смотрите также

computer; MP system; MPP; multiprocessor; parallel processing; SMP; SSP

parallel computing         

общая лексика

параллельные вычисления, параллельная обработка

Смотрите также

parallel processing

parallel code         

общая лексика

параллельный код

parallel program         

общая лексика

параллельная программа

программа, предназначенная для параллельного выполнения

Смотрите также

parallel processing; parallel software

parallel machine         

общая лексика

параллельный компьютер

машина параллельного действия

антоним

sequential machine

multicomputer         

Смотрите также

cluster

parallel programming         

общая лексика

параллельное программирование

методы и инструментарий разработки параллельного ПО

Смотрите также

parallel computer; parallel software

parallel transmission         
METHOD OF TRANSMITTING MULTIPLE BITS OF DATA SIMULTANEOUSLY
Parallel transmission; Parallel bus; Bit parallel; Parallel communicaiton; Parellel communication; Parallel communcations; Parallel communications

общая лексика

параллельная передача

передача символа, адреса или данных по нескольким раздельным линиям

Смотрите также

parallel port; transmission

parallel bars         
  • A gymnast performs on the parallel bars
APPARATUS USED IN MEN'S ARTISTIC GYMNASTICS
Parallel Bars; Gymnastics parallel bars; P bars; Parallel bars (gymnastics)

общая лексика

(параллельные) брусья

parallel bars         
  • A gymnast performs on the parallel bars
APPARATUS USED IN MEN'S ARTISTIC GYMNASTICS
Parallel Bars; Gymnastics parallel bars; P bars; Parallel bars (gymnastics)
parallel bars (параллельные) брусья

Определение

клавиатура
КЛАВИАТ'УРА, клавиатуры, ·муж. (·итал. claviatura) (спец.). Полный ряд клавишей в клавишных музыкальных инструментах (муз.).
| Совокупность клавишей в разного рода механизмах. Клавиатура пишущей машинки.

Википедия

Parallel computing

Parallel computing is a type of computation in which many calculations or processes are carried out simultaneously. Large problems can often be divided into smaller ones, which can then be solved at the same time. There are several different forms of parallel computing: bit-level, instruction-level, data, and task parallelism. Parallelism has long been employed in high-performance computing, but has gained broader interest due to the physical constraints preventing frequency scaling. As power consumption (and consequently heat generation) by computers has become a concern in recent years, parallel computing has become the dominant paradigm in computer architecture, mainly in the form of multi-core processors.

Parallel computing is closely related to concurrent computing—they are frequently used together, and often conflated, though the two are distinct: it is possible to have parallelism without concurrency, and concurrency without parallelism (such as multitasking by time-sharing on a single-core CPU). In parallel computing, a computational task is typically broken down into several, often many, very similar sub-tasks that can be processed independently and whose results are combined afterwards, upon completion. In contrast, in concurrent computing, the various processes often do not address related tasks; when they do, as is typical in distributed computing, the separate tasks may have a varied nature and often require some inter-process communication during execution.

Parallel computers can be roughly classified according to the level at which the hardware supports parallelism, with multi-core and multi-processor computers having multiple processing elements within a single machine, while clusters, MPPs, and grids use multiple computers to work on the same task. Specialized parallel computer architectures are sometimes used alongside traditional processors, for accelerating specific tasks.

In some cases parallelism is transparent to the programmer, such as in bit-level or instruction-level parallelism, but explicitly parallel algorithms, particularly those that use concurrency, are more difficult to write than sequential ones, because concurrency introduces several new classes of potential software bugs, of which race conditions are the most common. Communication and synchronization between the different subtasks are typically some of the greatest obstacles to getting optimal parallel program performance.

A theoretical upper bound on the speed-up of a single program as a result of parallelization is given by Amdahl's law, which states that it is limited by the fraction of time for which the parallelization can be utilised.

Как переводится parallel computer на Русский язык